Modeling the Development of Audiovisual Cue Integration in Speech Perception

نویسندگان

  • Laura M. Getz
  • Elke R. Nordeen
  • Sarah C. Vrabic
  • Joseph C. Toscano
چکیده

Adult speech perception is generally enhanced when information is provided from multiple modalities. In contrast, infants do not appear to benefit from combining auditory and visual speech information early in development. This is true despite the fact that both modalities are important to speech comprehension even at early stages of language acquisition. How then do listeners learn how to process auditory and visual information as part of a unified signal? In the auditory domain, statistical learning processes provide an excellent mechanism for acquiring phonological categories. Is this also true for the more complex problem of acquiring audiovisual correspondences, which require the learner to integrate information from multiple modalities? In this paper, we present simulations using Gaussian mixture models (GMMs) that learn cue weights and combine cues on the basis of their distributional statistics. First, we simulate the developmental process of acquiring phonological categories from auditory and visual cues, asking whether simple statistical learning approaches are sufficient for learning multi-modal representations. Second, we use this time course information to explain audiovisual speech perception in adult perceivers, including cases where auditory and visual input are mismatched. Overall, we find that domain-general statistical learning techniques allow us to model the developmental trajectory of audiovisual cue integration in speech, and in turn, allow us to better understand the mechanisms that give rise to unified percepts based on multiple cues.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Dynamic Modeling Approaches for Audiovisual Speech Perception and Multisensory Integration

Multimodal information including auditory, visual and even haptic information is integrated during speech perception. Articulatory information provided by a talker‘s face enhances speech intelligibility in congruent and temporally coincident signals, and produces a perceptual fusion (e.g. the ―McGurk effect‖) when the auditory and visual signals are incongruent. This paper focuses on promising ...

متن کامل

Audiovisual speech integration and lipreading in autism.

BACKGROUND During speech perception, the ability to integrate auditory and visual information causes speech to sound louder and be more intelligible, and leads to quicker processing. This integration is important in early language development, and also continues to affect speech comprehension throughout the lifespan. Previous research shows that individuals with autism have difficulty integrati...

متن کامل

Some behavioral and neurobiological constraints on theories of audiovisual speech integration: a review and suggestions for new directions.

Summerfield (1987) proposed several accounts of audiovisual speech perception, a field of research that has burgeoned in recent years. The proposed accounts included the integration of discrete phonetic features, vectors describing the values of independent acoustical and optical parameters, the filter function of the vocal tract, and articulatory dynamics of the vocal tract. The latter two ac...

متن کامل

Uncertainty and attention in audiovisual speech perception

This study deals with uncertainty and attention in audiovisual speech perception. Subjects were exposed to audiovisual stimuli where independent variable was bivalent: either blocked to one ear or uncertainty about in which ear the next stimulus would appear. The hypothesis, that uncertainty would inhibit audiovisual integration was motivated by an earlier study (Öhrström et al, 2011) but these...

متن کامل

Multisensory and sensorimotor interactions in speech perception

This research topic presents speech as a natural, well-learned, multisensory communication signal, processed by multiple mechanisms. Reflecting the general status of the field, most articles focus on audiovisual speech perception and many utilize the McGurk effect, which arises when discrepant visual and auditory speech stimuli are presented (McGurk and MacDonald, 1976). Tiippana (2014) argues ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره 7  شماره 

صفحات  -

تاریخ انتشار 2017